Single candidate optimizer: a novel optimization algorithm

نویسندگان

چکیده

Abstract Single-solution-based optimization algorithms have gained little to no attention by the research community, unlike population-based approaches. This paper proposes a novel algorithm, called Single Candidate Optimizer (SCO), that relies only on single candidate solution throughout whole process. The proposed algorithm implements unique set of equations effectively update position solution. To balance exploration and exploitation, SCO is integrated with two-phase strategy where updates its differently in each phase. effectiveness approach validated testing it thirty three classical benchmarking functions four real-world engineering problems. compared well-known algorithms, i.e., Particle Swarm Optimization, Grey Wolf Optimizer, Gravitational Search Algorithm recent high-performance algorithms: Equilibrium Archimedes Optimization Algorithm, Mayfly Salp Algorithm. According Friedman Wilcoxon rank-sum tests, can significantly outperform all other for majority investigated results achieved motivates design development new single-solution-based further improve performance. source code publicly available at: https://uk.mathworks.com/matlabcentral/fileexchange/116100-single-candidate-optimizer .

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Flying Squirrel Optimizer (FSO): A novel SI-based optimization algorithm for engineering problems

This paper provides a novel meta-heuristic optimization algorithm. The behaviors of flying squirrels in the nature are the main inspiration of this research. These behaviors include flying from tree to tree and walking on the ground or on a tree branch to find food. They also contact each other with chirp or squeak. This algorithm is named flying squirrel optimizer (FSO). Two main theories of m...

متن کامل

Neumann Optimizer: A Practical Optimization Algorithm for Deep Neural Networks

Progress in deep learning is slowed by the days or weeks it takes to train large models. The natural solution of using more hardware is limited by diminishing returns, and leads to inefficient use of additional resources. In this paper, we present a large batch, stochastic optimization algorithm that is both faster than widely used algorithms for fixed amounts of computation, and also scales up...

متن کامل

Neumann Optimizer: a Practical Optimization Algorithm for Deep Neural Networks

Progress in deep learning is slowed by the days or weeks it takes to train large models. The natural solution of using more hardware is limited by diminishing returns, and leads to inefficient use of additional resources. In this paper, we present a large batch, stochastic optimization algorithm that is both faster than widely used algorithms for fixed amounts of computation, and also scales up...

متن کامل

Neumann Optimizer: a Practical Optimization Algorithm for Deep Neural Networks

Progress in deep learning is slowed by the days or weeks it takes to train large models. The natural solution of using more hardware is limited by diminishing returns, and leads to inefficient use of additional resources. In this paper, we present a large batch, stochastic optimization algorithm that is both faster than widely used algorithms for fixed amounts of computation, and also scales up...

متن کامل

A NOVEL META-HEURISTIC ALGORITHM: TUG OF WAR OPTIMIZATION

This paper presents a novel population-based meta-heuristic algorithm inspired by the game of tug of war. Utilizing a sport metaphor the algorithm, denoted as Tug of War Optimization (TWO), considers each candidate solution as a team participating in a series of rope pulling competitions.  The  teams  exert  pulling  forces  on  each  other...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Evolutionary Intelligence

سال: 2022

ISSN: ['1864-5909', '1864-5917']

DOI: https://doi.org/10.1007/s12065-022-00762-7